Bounded Loss Classification∗

نویسنده

  • Jason D. M. Rennie
چکیده

Consider the problem of classification. Modern-day solutions have looked toward problem formulations where the search space is convex. Such formulations guarantee that a minimization of the objective function is found. But, in order to achieve that guarantee, such formulations treat outliers somewhat overzealously. Many classification objectives can be viewed as minimizing a loss funciton. For example, the Support Vector Machine minimizes the hinge loss, ∑ i lh(w T xi + b), where lh(z) = (1 − z)+. Logistic regression (LR) minimizes a similar loss, the logistic, ∑ i ll(w T xi + b), where ll(z) = − log 1 1+e−x . Many would not call these two losses “similar,” but they share important properties. One is that they are convex, e.g. lh(αz1 + (1− α)z2) ≥ αlh(z1) + (1− α)lh(z2). This is the reason that the objective has a unique minimum. Along with this property comes a less desirable property, that the loss for a single example is unbounded. In other words, an outlier can have an unbounded effect on the decision boundary. In practice, regularization is used to temper this effect, but it can produce negative effects. In summary, state-of-the-art classifiers utilize a convex loss function to achieve an objective with a unique minimum, but as a result, outliers can significantly effect the decision boundary. Motivation for current, state-of-the-art techniques was a long history of classification algorithms that used non-convex objective functions. For classification, one wishes to learn a decision boundary that will minimize the zero-one loss (lz(x) = θ(−wT x + b), θ is the heavaside function) on unseen examples drawn from the same distribution as the training examples. Many objectives minimized this or something very similar. Of course, such an optimization is riddled with local minima. Various techniques were developed to work around this problem, but none was as effective as the convex objectives that have been recently brought forward. We believe there is still hope in a more traditional objective, one that minimizes the zero-one loss. There are issues to be addressed—since the zero-one loss is not convex, how can we find a good solution? But, we feel that there are ways to sufficiently address this issue. Additionally, with the ∗Joint work with Tommi Jaakkola, [email protected]

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Estimation of Scale Parameter Under a Bounded Loss Function

     The quadratic loss function has been used by decision-theoretic statisticians and economists for many years.  In this paper  the estimation of scale parameter under a bounded loss function, which is adequate for assessing quality and quality improvement, is considered with restriction to the principles of invariance and risk unbiasedness. An implicit form of minimum risk scale equivariant ...

متن کامل

Risk premiums and certainty equivalents of loss-averse newsvendors of bounded utility

Loss-averse behavior makes the newsvendors avoid the losses more than seeking the probable gains as the losses have more psychological impact on the newsvendor than the gains. In economics and decision theory, the classical newsvendor models treat losses and gains equally likely, by disregarding the expected utility when the newsvendor is loss-averse. Moreover, the use of unbounded utility to m...

متن کامل

Minimax Estimator of a Lower Bounded Parameter of a Discrete Distribution under a Squared Log Error Loss Function

The problem of estimating the parameter ?, when it is restricted to an interval of the form , in a class of discrete distributions, including Binomial Negative Binomial discrete Weibull and etc., is considered. We give necessary and sufficient conditions for which the Bayes estimator of with respect to a two points boundary supported prior is minimax under squared log error loss function....

متن کامل

Estimating a Bounded Normal Mean Relative to Squared Error Loss Function

Let be a random sample from a normal distribution with unknown mean and known variance The usual estimator of the mean, i.e., sample mean is the maximum likelihood estimator which under squared error loss function is minimax and admissible estimator. In many practical situations, is known in advance to lie in an interval, say for some In this case, the maximum likelihood estimator...

متن کامل

Estimation of Lower Bounded Scale Parameter of Rescaled F-distribution under Entropy Loss Function

We consider the problem of estimating the scale parameter &beta of a rescaled F-distribution when &beta has a lower bounded constraint of the form &beta&gea, under the entropy loss function. An admissible minimax estimator of the scale parameter &beta, which is the pointwise limit of a sequence of Bayes estimators, is given. Also in the class of truncated linear estimators, the admissible estim...

متن کامل

Improved Estimation in Rayleigh type-II Censored Data under a Bounded Loss Utilizing a Point Guess Value

‎The problem of shrinkage testimation (test-estimation) for the Rayleigh scale‎ ‎parameter θ based on censored samples under the reflected‎ ‎gamma loss function is considered‎. We obtain the minimum risk‎ ‎estimator among a subclass and compute its risk‎. ‎A shrinkage‎ ‎testimator based on acceptance or rejection of a null hypothesis&lr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2003